Goto

Collaborating Authors

 perceptual quality


Resource-constrained image generation and visual understanding: an interview with Aniket Roy

AIHub

In the latest in our series of interviews meeting the AAAI/SIGAI Doctoral Consortium participants, we caught up with Aniket Roy to find out more about his research on generative models for computer vision tasks. Tell us a bit about your PhD - where did you study, and what was the topic of your research? I recently completed my PhD in Computer Science at Johns Hopkins University, where I worked under the supervision of Bloomberg Distinguished Professor Rama Chellappa. My research primarily focused on developing methods for resource-constrained image generation and visual understanding. In particular, I explored how modern generative models can be adapted to operate efficiently while maintaining strong performance.


Lossy Image Compression with Conditional Diffusion Models

Neural Information Processing Systems

In contrast to V AE-based neural compression, where the (mean) decoder is a deterministic neural network, our decoder is a conditional diffusion model. Our approach thus introduces an additional "content" latent variable on which the reverse diffusion process









Looks TooGoodToBeTrue: AnInformation-TheoreticAnalysisofHallucinations inGenerativeRestorationModels

Neural Information Processing Systems

The pursuit of high perceptual quality in image restoration has driven the development of revolutionary generative models, capable of producing results often visually indistinguishable from real data. However, as their perceptual quality continues toimprove, these models also exhibit agrowing tendencytogenerate hallucinations -realistic-looking details that do not exist in the ground truth images. Hallucinations in these models create uncertainty about their reliability, raising major concerns about their practical application.